Facial animation based on context-dependent visemes
نویسندگان
چکیده
This paper presents a novel approach for the generation of realistic speech synchronized 3D facial animation that copes with anticipatory and perseveratory coarticulation. The methodology is based on the measurement of 3D trajectories of fiduciary points marked on the face of a real speaker during the speech production of CVCV non-sense words. The trajectories are measured from standard video sequences using stereo vision photogrammetric techniques. The first stationary point of each trajectory associated with a phonetic segment is selected as its articulatory target. By clustering according to geometric similarity all articulatory targets of a same segment in different phonetic contexts, a set of phonetic context-dependent visemes accounting for coarticulation is identified. These visemes are then used to drive a set of geometric transformation/deformation models that reproduce the rotation and translation of the temporomandibular joint on the 3D virtual face, as well as the behavior of the lips, such as protrusion, and opening width and height of the natural articulation. This approach is being used to generate 3D speech synchronized animation from both natural and synthetic speech generated by a text-to-speech synthesizer. r 2006 Elsevier Ltd. All rights reserved.
منابع مشابه
Benchmarking Speech Synchronized Facial Animation Based on Context-Dependent Visemes
In this paper we evaluate the effectiveness in conveying speech information of a speech synchronized facial animation system based on context-dependent visemes. The evaluation procedure is based on an oral speech intelligibility test conducted with, and without, supplementary visual information provided by a real and a virtual speaker. Three situations (audio-only, audio+video and audio+animati...
متن کامل3D Facial Animation for Mobile Devices
This article presents the implementation of a 3D facial animation system for mobile devices. Due to the large processing and memory requirements for this type of application, its use on mobile devices was not possible until recently. Currently, however, with the increasing development of powerful hardware and with the spread of cellular telephony, 3D applications for these devices have become e...
متن کاملThe Development of a Brazilian Talking Head
This paper describes partial results of a research, in progress at the School of Electrical and Computer Engineering of the State University of Campinas, aimed at developing a realistic three-dimensional Brazilian Talking Head. Through an extensive analysis of a video-audio linguistic corpus, a set of 29 phonetic context-dependent visemes (22 consonantal plus 7 vocalic visemes), that accommodat...
متن کاملLearning-Based Facial Animation
This thesis proposes a novel approach for automated 3D speech animation from audio. An end-to-end system is presented which undergoes three principal phases. In the acquisition phase, dynamic articulation motions are recorded and amended. The learning phase studies the correlation of these motions in their phonetic context in order to understand the visual nature of speech. Finally, for the syn...
متن کاملVisual analysis of viseme dynamics
Face to face dialogue is the most natural mode of communication between humans. The combination of human visual perception of expression and perception in changes in intonation provides semantic information that communicates idea, feelings and concepts. The realistic modelling of speech movements, through automatic facial animation, and maintaining audio-visual coherence is still a challenge in...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- Computers & Graphics
دوره 30 شماره
صفحات -
تاریخ انتشار 2006